Identifying Mixtures of Mixtures Using Bayesian Estimation
نویسندگان
چکیده
The use of a finite mixture of normal distributions in model-based clustering allows us to capture non-Gaussian data clusters. However, identifying the clusters from the normal components is challenging and in general either achieved by imposing constraints on the model or by using post-processing procedures. Within the Bayesian framework, we propose a different approach based on sparse finite mixtures to achieve identifiability. We specify a hierarchical prior, where the hyperparameters are carefully selected such that they are reflective of the cluster structure aimed at. In addition, this prior allows us to estimate the model using standard MCMC sampling methods. In combination with a post-processing approach which resolves the label switching issue and results in an identified model, our approach allows us to simultaneously (1) determine the number of clusters, (2) flexibly approximate the cluster distributions in a semiparametric way using finite mixtures of normals and (3) identify cluster-specific parameters and classify observations. The proposed approach is illustrated in two simulation studies and on benchmark datasets. Supplementary materials for this article are available online.
منابع مشابه
A SIMPLE MODEL FOR THE ESTIMATION OF DIELECTRIC CONSTANTS OF BINARY SOLVENT MIXTURES
A simple and reliable method for quick estimation of the dielectric constant of a binary solvent mixture is proposed. The validity of the proposed method has been tested for a broad range of binary solvent mixtures
متن کاملConvergence Rates in Nonparametric Bayesian Density Estimation
We consider Bayesian density estimation for compactly supported densities using Bernstein mixtures of beta-densities equipped with a Dirichlet prior on the distribution function. We derive the rate of convergence for α-smooth densities for 0 < α ≤ 2 and show that a faster rate of convergence can be obtained by using fewer terms in the mixtures than proposed before. The Bayesian procedure adapts...
متن کاملBayesian estimation of mixtures with dynamic transitions and known component parameters
Probabilistic mixtures provide flexible “universal” approximation of probability density functions. Their wide use is enabled by the availability of a range of efficient estimation algorithms. Among them, quasi-Bayesian estimation plays a prominent role as it runs “naturally” in one-pass mode. This is important in on-line applications and/or extensive databases. It even copes with dynamic natur...
متن کاملLearning mixtures of polynomials from data using B-spline interpolation
Hybrid Bayesian networks efficiently encode a joint probability distribution over a set of continuous and discrete variables. Several approaches have been recently proposed for working with hybrid Bayesian networks, e.g., mixtures of truncated basis functions, mixtures of truncated exponentials or mixtures of polynomials (MoPs). We present a method for learning MoP approximations of probability...
متن کاملA Mixture of Coalesced Generalized Hyperbolic Distributions
A mixture of coalesced generalized hyperbolic distributions (GHDs) is developed by joining a finite mixture of generalized hyperbolic distributions with a novel mixture of multiple scaled generalized hyperbolic distributions (MSGHDs). After detailing the development of the mixture of MSGHDs, which arises via implementation of a multidimensional weight function, the density of our coalesced dist...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره 26 شماره
صفحات -
تاریخ انتشار 2017